Generalized Conditional Gradient for Sparse Estimation
نویسندگان
چکیده
Sparsity is an important modeling tool that expands the applicability of convex formulations for data analysis, however it also creates significant challenges for efficient algorithm design. In this paper we investigate the generalized conditional gradient (GCG) algorithm for solving sparse optimization problems—demonstrating that, with some enhancements, it can provide a more efficient alternative to current state of the art approaches. After studying the convergence properties of GCG for general convex composite problems, we develop efficient methods for evaluating polar operators, a subroutine that is required in each GCG iteration. In particular, we show how the polar operator can be efficiently evaluated in learning low-rank matrices, instantiated with detailed examples on matrix completion and dictionary learning. A further improvement is achieved by interleaving GCG with fixedrank local subspace optimization. A series of experiments on matrix completion, multi-class classification, and multi-view dictionary learning shows that the proposed method can significantly reduce the training cost of current alternatives.
منابع مشابه
Equivalence of a Generalized Conditional Gradient Method and the Method of Surrogate Functionals
This article combines techniques from two fields of applied mathematics: optimization theory and inverse problems. We investigate the equivalence of the classical conditional gradient method and the surrogate method, which has been recently proposed for solving inverse problems. The method of surrogate functionals aims at the solution of non-quadratic minimization problems where the solution is...
متن کاملLarge-scale Inversion of Magnetic Data Using Golub-Kahan Bidiagonalization with Truncated Generalized Cross Validation for Regularization Parameter Estimation
In this paper a fast method for large-scale sparse inversion of magnetic data is considered. The L1-norm stabilizer is used to generate models with sharp and distinct interfaces. To deal with the non-linearity introduced by the L1-norm, a model-space iteratively reweighted least squares algorithm is used. The original model matrix is factorized using the Golub-Kahan bidiagonalization that proje...
متن کاملEfficient Generalized Conditional Gradient with Gradient Sliding for Composite Optimization
Generalized conditional gradient method has regained increasing research interest as an alternative to another popular proximal gradient method for sparse optimization problems. For particular tasks, its low computation cost of linear subproblem evaluation on each iteration leads to superior practical performance. However, the inferior iteration complexity incurs excess number of gradient evalu...
متن کاملA generalized conditional gradient method for nonlinear operator equations
The intention of this paper is to show the applicability of a generalized conditional gradient method for the minimization of Tikhonov-type functionals, which occur in the regularization of nonlinear inverse problems with sparsity constraints. We consider functionals of Tikhonov type where the usual quadratic penalty term is replaced by the pth power of a weighted p-norm. First of all, we analy...
متن کاملPolar Operators for Structured Sparse Estimation
Structured sparse estimation has become an important technique in many areas of data analysis. Unfortunately, these estimators normally create computational difficulties that entail sophisticated algorithms. Our first contribution is to uncover a rich class of structured sparse regularizers whose polar operator can be evaluated efficiently. With such an operator, a simple conditional gradient m...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Journal of Machine Learning Research
دوره 18 شماره
صفحات -
تاریخ انتشار 2017